#LLM optimization19/10/2025
BitDistill: Microsoft’s 1.58‑Bit Pipeline Cuts Memory by 10× and Speeds CPU by ~2.65×
'Microsoft Research presents BitNet Distillation: a three‑stage pipeline that converts FP16 models to 1.58‑bit students, delivering up to 10× memory reduction and ~2.65× CPU speedup while preserving accuracy.'